Learning output kernels for multi-task problems
نویسندگان
چکیده
منابع مشابه
Learning output kernels for multi-task problems
Simultaneously solving multiple related learning tasks is beneficial under a variety of circumstances, but the prior knowledge necessary to correctly model task relationships is rarely available in practice. In this paper, we develop a novel kernel-based multi-task learning technique that automatically reveals structural inter-task relationships. Building over the framework of output kernel lea...
متن کاملKernels for Multi--task Learning
This paper provides a foundation for multi–task learning using reproducing kernel Hilbert spaces of vector–valued functions. In this setting, the kernel is a matrix–valued function. Some explicit examples will be described which go beyond our earlier results in [7]. In particular, we characterize classes of matrix– valued kernels which are linear and are of the dot product or the translation in...
متن کاملUniversal Kernels for Multi-Task Learning
In this paper we are concerned with reproducing kernel Hilbert spaces HK of functions from an input space into a Hilbert space Y, an environment appropriate for multi-task learning. The reproducing kernel K associated to HK has its values as operators on Y. Our primary goal here is to derive conditions which ensure that the kernel K is universal. This means that on every compact subset of the i...
متن کاملMulti-task Learning for Structured Output Prediction
Facial landmark detection is an important step for many perception tasks. In this paper, we address facial landmark detection as a structured output regression problem, where we exploit the strong dependencies that lie between the facial landmarks. For this, we propose a generic multi-task regression framework for structured output problems. The learning of the output structure is achieved thro...
متن کاملMulti-task and Lifelong Learning of Kernels
We consider a problem of learning kernels for use in SVM classification in the multi-task and lifelong scenarios and provide generalization bounds on the error of a large margin classifier. Our results show that, under mild conditions on the family of kernels used for learning, solving several related tasks simultaneously is beneficial over single task learning. In particular, as the number of ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Neurocomputing
سال: 2013
ISSN: 0925-2312
DOI: 10.1016/j.neucom.2013.02.024